2024-06-25 17:57:14.AIbase.9.8k
Apple Launches All-in-One Visual Model 4M-21 Capable of Handling 21 Different Modalities
Apple and researchers from the Swiss Federal Institute of Technology (EPFL) have jointly developed a single model that can handle arbitrary-to-arbitrary modalities. This model can be trained on dozens of highly diverse modalities and perform joint training on large-scale multimodal datasets and text corpora. Named 4M-21, it is trained on 21 different modalities, completing at least three times more tasks than existing models without losing performance.